Next Article in Journal
Experimental and Numerical Study of Flexural Stiffness Performance of Ultra-Thin, Prefabricated, and Laminated Slab Base Slabs
Next Article in Special Issue
Time Series Data Modeling Using Advanced Machine Learning and AutoML
Previous Article in Journal
Conservation Hub: The Added Value of the Whale-Watching Industry
Previous Article in Special Issue
Smart Water Resource Management Using Artificial Intelligence—A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Ensemble Machine Learning Technique for Detection of Abnormalities in Knee Movement Sustainability

1
Computer Science and Engineering Department, Thapar Institute of Engineering and Technology, Patiala 147004, India
2
School of Computer Science, University of Petroleum and Energy Studies, Dehradun 248007, India
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(20), 13464; https://doi.org/10.3390/su142013464
Submission received: 29 July 2022 / Revised: 9 October 2022 / Accepted: 17 October 2022 / Published: 19 October 2022

Abstract

:
The purpose of this study was to determine electromyographically if there are significant differences in the movement associated with the knee muscle, gait, leg extension from a sitting position and flexion of the leg upwards for regular and abnormal sEMG data. Surface electromyography (sEMG) data were obtained from the lower limbs of 22 people during three different exercises: sitting, standing, and walking (11 with and 11 without knee abnormality). Participants with a knee deformity took longer to finish the task than the healthy subjects. The sEMG signal duration of patients with abnormalities was longer than that of healthy patients, resulting in an imbalance in the obtained sEMG signal data. As a result of the data’s bias towards the majority class, developing a classification model for automated analysis of such sEMG signals is arduous. The sEMG collected data were denoised and filtered, followed by the extraction of time-domain characteristics. Machine learning methods were then used for predicting the three distinct movements (sitting, standing, and walking) associated with electrical impulses for normal and abnormal sets. Different anomaly detection techniques were also used for detecting occurrences in the sEMG signals that differed considerably from the majority of data and were hence used for enhancing the performance of our model. The iforest anomaly detection technique presented in this work can achieve 98.5% accuracy on the light gradient boosting machine algorithm, surpassing the previous results which claimed a maximum accuracy of 92.5% and 91%, improving accuracy by 6–7% for classification of knee abnormality using machine learning.

1. Introduction

Walking, running, or climbing stairs might be challenging if you have discomfort due to joint pain. Pain can arise as a result of a trauma or an injury. It may also start for no apparent cause at times. One study found that one in three people aged between 18 and 64 years have arthritis or joint symptoms due to an underlying injury or condition, such as osteoarthritis of the knee [1]. X-ray [2] or magnetic resonance imaging (MRI) methods [3] are used in the clinic to diagnose such abnormalities. The X-ray approach collects and assesses bone condition, whereas MRI gives comprehensive information about knee anatomies, such as cartilage, ligaments, and tendons. MRI is an effective diagnostic tool, but it is also more expensive. Wearable sensors, such as an electromyography (EMG), accelerometers, and gyro meters, and visual sensors such as imaging cameras can also be used to identify knee problems.
Depending on the parts of the human body supported by an exoskeleton, they can be divided into upper limb exoskeleton, lower limb exoskeleton (lees), whole body exoskeleton, and specific joint exoskeleton [4,5]. These exoskeleton systems are divided into three categories based on their various target applications and users, which are gait rehabilitation, human movement, and enhancement of human force [6].
Computer vision has applications across all areas of engineering, including exoskeletons. The exoskeleton obtains information about its environment from an RGB-D camera and extracts characteristics of the soil surface that could affect its gait. It then makes decisions according to the features of the environment, the state of the environment, and safety restrictions. It also provides the correct length and height of the walk to the parameterized gait planning model to assist the user in walking [7].
Sun et al. present the generation of exoskeleton mobility; several signal acquisition types can be applied, each focusing on specific areas and movements of the human body [8]. The current signal acquisition technologies include computer vision, electroencephalography (EEG), electromyography (EMG), and inertial sensors. These methods have been developed over the past several years. In addition, the detection and control method, stability of the modeling, and comfort of an exoskeleton of the lower limb will also impact its performance [8].
Calle et al. discussed the integration of inertial sensors into a lower limb robotic exoskeleton [9]. They present a novel platform based on Imocap-GIS inertial sensors with a motion-assistance exoskeleton. It generates joint movements via Maxon motors and harmonic drive gearboxes.
Tang et al. present an extensive study on a wearable lower-limb exoskeleton [10]. They classified the wearable device according to the power source and the working principle. They also compared and analyzed the design idea, wearing mode, material, and performance of different types of lower-limb exoskeletons.
Taylor et al. demonstrate real-time muscle length tracking in an in vivo turkey model via chronically implanted magnetic beads while investigating accuracy, biocompatibility, and long-term implant stability [11].
Botter et al. combined ultrasound (US) imaging and multichannel surface EMG to compare their detection sensitivity to fasciculations occurring in different muscle regions and to investigate the effect of EMG electrodes’ configuration on their agreement [12]. Monopolar surface EMGs were collected from medial gastrocnemius and soleus with an array of 32 electrodes.
EMG sensors recognize motions in advance while detecting signal changes faster, making them better at researching neuromusculoskeletal diseases. The signal obtained by the EMG sensors is a biological signal which measures the electrical activity produced by the skeletal muscles. Surface electromyogram (sEMG) and intramuscular EMG (iEMG) techniques are important methods to acquire EMG signals [13]. sEMG has an advantage over iEMG in that the electrodes may be placed without causing discomfort or requiring medical supervision, and the risk of infection is minimum.
Gui et al. proposed an adaptive estimator based on the EMG to obtain and update the EMG pair model without calibrations and recalibrations [14]. Simulation and experiments indicate that the proposed estimator can adaptively predict the subject’s active joint couple and provide precise control of exoskeleton movement.
Kang et al. developed a neural network gait and slope estimator using electromyogram (EMG) and mechanical sensor signals [15]. The results of four healthy and two elderly subjects show that the EMG approach can reduce the error rate by 14.8% compared to models using only mechanical sensors.
Long-term monitoring with surface electrodes is simpler than with iEMG needles. As Kanoga et al. have shown, the location of the sEMG sensors has a considerable impact on the collected signal and its subsequent analysis and identification using the calculation method [16]. sEMG sensors can collect data during daily human activities, such as sitting, climbing, walking, and standing, which can then be used to recognize actions, detect anomalies, etc. The goniometer was connected out of the knee joint, and sEMG data were collected for four muscles: rectus femoris (RF), biceps femoris (BF), vastus medialis (VM), semitendinosus (ST), and semimembranosus (SM) (Figure 1).

1.1. Motivation

Knee pain has become one of the most common health issues among seniors. Wounds, repeated stress on the joint, aging, and underlying conditions such as arthritis are the most common causes of knee pain [17]. The knee joint is one of the most complex joints in the human body, ensuring leg movement and body stability and acting as a damper [18]. The knee joint is made up of different bones, tendons, muscles, cartilage, ligaments, and fluids.
During the past decade, researchers have focused on categorizing sEMG signals from the upper limb. The sEMGs obtained from the lower limb are more difficult to interpret because they are influenced by overlapping muscles [17,18]. Various signal processing approaches [19,20] have been used to study the categorization of sEMG data from the lower leg. Chen et al. used a deep-neural-network to predict lower-limb extension/flexion joint angles from sEMG signals [21]. Morbidoni et al. used an in-depth learning method to classify the walking step using sEMG data [22]. Bonato et al. examined the depletion of quadriceps and hamstring muscles using sEMG signals [23]. The classification of walking patterns and the use of computational approaches are divided into normal and gait disorders [24].

1.2. Contributions

This paper focuses on identifying abnormality in movements from imbalanced sEMG data and classifying six different activities (three normal and three abnormal) for standing, sitting, and gait positions. We thus present a performance comparison of various prediction classifiers for detecting knee abnormalities from sEMG signals in which classifiers are applied on the datasets consisting of data with and without anomalies in using multiple anomaly detection techniques. This paper aims to enhance the ability to detect knee movement abnormalities and enable healthcare for a more accurate diagnosis. To the author’s knowledge, no study solves the problem of identifying knee abnormalities using anomalous detection methods. The major contributions are described as follows:
  • Anomaly detection techniques such as iforest, KNN, and lof have been used to remove abnormal data.
  • Enhancement of classifier’s accuracy using boosting algorithms.

1.3. Organization

In this paper, Section 2 describes our proposed model’s datasets, methodology, and workflow. Section 3 discusses the model evaluation parameters. Section 4 defines the performance and provides a comparative analysis between different metrics. Finally, the paper is concluded in Section 5.

2. Materials and Methods

2.1. Dataset and Its Features

The sEMG and joint angles were derived from the publicly available dataset at UCI machine learning repository [25]. The data includes sEMG signals from the lower limbs of 22 adults, 11 of whom are in good health and 11 of whom have known knee problems. Participants were in good health with no history of injury or knee discomfort. One abnormal subject of the knee suffered a sciatic nerve lesion, six subjects suffered an anterior cruciate ligament (ACL) lesion, and the other four suffered a meniscus lesion. Dataset description and feature information have been provided in Table 1 and Table 2, respectively. These data were collected using a Biometrics Ltd. DataLog MWX8 and a direction finder. Each person performed three distinct tasks: walking, bending the leg upward, and lengthening the leg from a sitting posture. The goniometer was connected out of the knee joint, and sEMG data was collected for four muscles: vastus medialis (VM), rectus femoris (RF), semitendinosus (ST), and biceps femoris (BF) (Figure 1).
The person’s afflicted limb with aberrant knee and left leg healthy participants were chosen to receive the signal. The data were collected at a sampling frequency of 1000 Hz with a resolution of 14 bits. The sEMG signals had already been filtered through a band-pass filter ranging from 20 Hz to 460 Hz. The data gathered do not have a signal corresponding to the transition steps, such as standing, sitting, walking, standing, etc. In Table 3, movements labeled as 0 refer to normal walking, 1 for normal sitting, 2 for normal standing, 3 for abnormal walking, 4 for abnormal sitting, and 5 for abnormal standing, respectively. The data were transmitted in real-time through Bluetooth from the MWX8 gadget to the PC.

Statistical Analysis of Dataset

Table 4 and Figure 2 show the statistical analysis of all the datasets. The standard deviation of FX is the highest, and BF is the lowest in all the datasets. EMG signal plots can be seen before data preprocessing in Figure 2a for the four channels of a normal person and in Figure 2b for an abnormal person. EMG signal plots can be seen after signal preprocessing, in which we rectify and transform the filtered sEMG values we can see in Figure 2c for a normal person and in Figure 2d for an abnormal person. After normalizing, we achieve the plots as shown in Figure 2e for a normal person and in Figure 2f for an abnormal person.

2.2. Approach

The approach is described in Figure 3. Firstly, the raw experimental data of the lower limb is taken from the UCI machine learning repository. The dataset consists of data for three different activities (sitting, standing, and walking) for 11 abnormal and 11 normal individuals labeled as Abn. And Nor. In Table 4.
The data is pre-filtered with a band-pass filter that ranges from 25 Hz to 465 Hz. The signal data is later demodulated using Hilbert’s transformation. In case of mechanical failure, the gathered vibration signals are usually modified. As such, signal demodulation can separate the carrier component from the modulation component. In addition, the default features are still hidden in the modulation component. Thus, implementation of signal demodulation during signal preprocessing is essential for precise results. The min/max normalization method was applied to normalize the samples with features ranging from 0 to 1, as shown in Figure 3. The dataset is then divided into two groups:
  • Dataset consisting of anomalies
  • Datasets produced as a result of the use of anomaly detection techniques, namely iforest, KNN, and lof.
Following the generation of these datasets, we use machine learning classifiers that comprise of models:
  • Light gradient boosting machine
  • Extra trees classifier
  • Random forest classifier
  • Extreme gradient boosting
  • Decision tree classifier

2.2.1. Anomaly Detection Techniques

Anomaly detection techniques (also known as outlier analysis) is a stage in data mining that finds data points, events, and/or observations that deviate from the normal behavior. Anomalous analytics may indicate major problems, such as a technical flaw or potential scenarios. We analyze the sEMG signals and remove outliers from them using the following techniques:
  • iforest: The Isolation Forest is an unsupervised learning algorithm to detect abnormalities that operate on the principle of isolating abnormalities rather than the most common staining profiling techniques [26]. The decision tree algorithm backs the secluded forest. It separates aberrant values by randomly selecting a character from the given set of characteristics and then randomly selecting a value divided between the max and min values. This random distribution of traits will lead to shorter tree trajectories for anomalous data points, differentiating them from the rest of the data. Generally, the first step in detecting an anomaly is to profile what is “normal” and then flag anything that cannot be considered typical as abnormal. However, the forest isolation algorithm does not function on this principle; it first defines the “normal” behavior and calculates the distances per point. As expected, the isolation forest functions instead by isolating anomalies by explicitly isolating abnormal issues in the dataset. The isolation forest algorithm assumes that specific abnormalities and observations should make it easier to identify them. The isolated forest uses a series of trees to isolate anomalies for the given data points [26]. The isolation forest recursively generates scores across the dataset by randomly selecting a feature and then a split value. The monsters likely need fewer random partitions than the “normal” points of the dataset. The faults will be the dots with a smaller path into the tree and the crossed edges of the root node.
  • KNN (k-Nearest Neighbors Detector): The unsupervised k-nearest-neighbor global 201 anomaly detection technique is an easy way to detect anomalies and not be confused with the k-nearest classification [27]. As the name suggests, it focuses on global anomalies and cannot see local ones. First, the closest k-neighbors must be found for every record in the data set. Afterward, an anomaly score is calculated using those neighbors [27]. On the other hand, two possibilities have been suggested: either the distance to the nearest kth-neighbor is used (only one), or the average length to all the nearest k-neighbors is calculated. In the subsequent paragraphs, we refer to the first method as kth-NN and the second as k-NN.
    In practice, the k-NN method is usually preferred. However, the absolute value of the score depends a lot on the data set itself, the number of dimensions, and the standardization [27]. Therefore, it is not easy to choose an appropriate threshold, if necessary. Naturally, selecting the parameter k is essential to the results. If too low, the density estimate for the records could not be reliable.
    In addition, the density estimate may be too coarse if it is too large. Usually, k should be in the range 10 < k < 50. In classification, a suitable k can be determined, for example, using cross-validation. Unfortunately, there is no such technique for detecting unsupervised abnormalities due to missing tags. For this reason, we later evaluate many different values for k and mean to obtain a fair assessment when comparing algorithms.
  • LOF (Local Outlier Factor): The LOF uses the identification of density-based outliers to identify local outliers or sites that are outliers relative to their immediate environment rather than the overall distribution of data [28]. The higher the LOF value of an observation, the more astonishing it becomes. The number of regarded neighbors is usually fixed to be: (1) less than the maximum number of samples close by which may be local outliers and (2) exceeds the minimum number of samples that a cluster must contain for other samples to be local outliers [28].

2.2.2. Feature Importance

Techniques that score each feature according to their usefulness in predicting a target variable are known as feature importance. Statistical correlation scores, coefficients derived from linear models, decision trees, and permutation significance scores are a few examples of trait importance scores. Feature significance scores are essential for predictive modeling projects. They provide an overview of the data and information on the template. A foundation for a reduction in dimensions and selection of characteristics can improve the effectiveness of a predictive model on the problem.
The importance of the feature is computed using Equation (1). After five different runs, the average weight given to each feature is described in Table 5 and their average weight is used for ranking the features. It is found that FX has the highest ranking and RF has the lowest ranking.
R a n k = i = 1 C f i   l o g ( f i )
where, C is the unique labels in the dataset, f i is the frequency if label i at a node.

2.2.3. Computational Classifiers

The dataset was divided into 75/25 Train-Test halves. The resulting datasets are subsequently trained on a variety of machine learning classifiers. The results are compared between the top five performing classifiers:
  • Light gradient boosting machine
  • Extra trees classifier
  • Random forest classifier
  • Extreme gradient boosting
  • Decision tree classifier
  • LightGBM: This improves gradient amplification by incorporating automatic feature selection and focusing on amplification instances with higher gradients [29]. This will lead to much faster workout times and greater predictive efficiency. It is known to be a very effective computational algorithm and rapid treatment method. As the trees of other algorithms grow horizontally, LightGBM grows vertically, in the direction of the leaves, while different algorithms grow in level. LightGBM picks the sheet with the highest loss to develop. When culturing the same sheet, it will decrease the loss more than a leveling method. Decision trees are used to improve model performance while saving memory. Quality and scalability are at the heart of the development approach. It uses two new techniques: one-sided gradient sampling and proprietary feature grouping (EFB). Fill in the histogram algorithm gaps in most GBDT (gradient boosting decision tree) frameworks [29]. The LightGBM algorithm’s characteristics are produced by the gradient-based one side sampling technique (GOSS) and the exclusive feature bundling technique (EFB). They work together to make the model work and give it an edge over the competition for GBDT systems.
  • Extreme gradient boosting: Unlike traditional gradient boosting, XGBoost uses its tree-building technique where similarity score and gain determine optimal node divisions [30]. Residual value: real value (observed)-predicted value. The probability of an event calculated in an earlier stage is called an earlier probability. For each finding, the initial probability is considered to be 0.5, which is used to create the first tree. The process map shows that the previous probability is reconstructed for all future trees based on the original predictions and the predictions of all previous trees. Lambda is a setting for equalization. The increase in lambda diminishes the influence of small leaves (those with few observations) disproportionately while having a moderate impact on large leaves (those with many observations) [30]. The optimum division for the tree is then determined by selecting the division knot with the greatest gain.
  • Random Forest: A random forest is a holistic approach that uses multiple decision trees [31]. It is also known as bootstrapping and aggregation, often called bagging, to perform regression and filing tasks. The simple hypothesis is that instead of relying on individual decision trees, the outcome is determined by combining multiple decision trees. Random Forest uses a variety of decision trees as a primary learning approach. The dataset is randomly sampled for line and feature sampling, resulting in sample data sets for each model. This section is called bootstrap. Random logging is a method of bagging and not a method of stimulation. Trees in random forests grow next to each other. Growing trees are not related. A randomized forest is a meta-marker (it incorporates the results of several predictive models) that groups together many decision trees with some helpful changes. It is one of the most precise machine learning technologies available on the market. On many datasets, this is a very reliable grader. It works effectively on big datasets [31]. It can handle tens of thousands of input variables without removing them. It identifies the importance of a large number of factors in categorization.
  • Extra Trees: This module presents a meta-estimator that uses the mean to improve statistical accuracy [32]. It combines a range of randomized decision trees. Extra trees is an intuitive learning approach that combines the predictions of multiple decision trees into a single prediction (the random forest model previously discussed). It may produce similar or better results than the random forest technique because it uses a simpler algorithm to create the entire decision trees. Following the standard upper-level approach, the extra-trees algorithm produces an arbitrary decision or a set of regression trees. It differs from earlier tree-based methods by randomly dividing nodes and generating trees using the whole learning sample (rather than a bootstrap replication) [32].
  • Decision Tree: Decision tree learning, also known as decision tree induction, is one of the statistical modeling methods used in analytics, data processing, and machine learning [33]. It uses a decision tree from assumptions about an item to predictions about the article’s target value. Classification trees represent class tags, and branches represent combinations of features that match specific class tags. A decision tree is a graphic and explicit depiction of the options and decision making that can be used in decision-making procedures. A decision tree is a data model for exploring the data. The decision tree learning approach is often used in data mining. The objective is to build a model that can predict the value of a target variable based on a set of input variables. The decision tree method is a supervised learning approach that can generate continuous and categoric production variables.

3. Model Evaluation

There are many classification metrics for evaluation of our model. In this study, we assess the performance on (1) Matthews correlation coefficient (MCC), (2) F1 Score (F1), (3) accuracy, (4) precision, and (5) AUC.

3.1. Matthews Correlation Coefficient (MCC)

The MCC formula takes into account all of the cells in the confusion matrix.
MCC = TP * TN FP * FN ( TP   +   FP ) ( TP   +   FN ) ( TN   +   FP ) ( TN   +   FN )
Similar to the correlation coefficient, MCC has a range of values ranging from −1 to +1. A model with a score of +1 is considered ideal, whereas a model with a score of −1 is considered bad. This feature is one of the most important benefits of MCC since it leads to simple interpretation.

3.2. F1 Score

The F or F score measurement is a measurement of test accuracy. It is calculated with the accuracy and recalculation values from the test. Accuracy is the number of actual positive outcomes divided by the total number of positive outcomes, including those incorrectly identified. The recall is the number of true positive results divided by the total number of samples which should have been positive.
F 1 = 2 * TP 2 * TP +   FP + FN

3.3. Precision

Precision, sometimes referred to as the positive predictive value, is computed as,
Precision = TP TP + FP
Precision may produce a biased result since the formula lacks FN and TN, especially for unbalanced classes.

3.4. Accuracy

Accuracy is a criterion for classifying models. Informally, precision is the percentage of accurate predictions from our model. Formal accuracy is defined by:
Accuracy = TP + TN TP + TN + FP + FN
where TP denotes true positives, TN denotes true negatives, FP denotes false positives, and FN denotes false negatives.

3.5. AUC

AUC is an abbreviation for “Area under the ROC Curve.” AUC measures the whole two-dimensional area beneath the entire ROC curve from (0,0) to (1,1). AUC is a metric that aggregates performance across all categorization criteria. AUC may be interpreted as the likelihood that the model rates a random positive case higher than a random negative example.

4. Result Analysis

Here, we examine the prediction outcomes of five machine learning classifiers on the training, testing, and validation dataset. The performance of each model is shown in Table 6. All the prediction models have been trained on default parameters and evaluated on MCC, accuracy, F1, precision, and AUC. The dataset has a small number of features but a large number of observation values. The K-fold cross-validation technique is used to assess the robustness of the best prediction method.

4.1. Model Performance Comparison

After training models on different datasets, it is observed that the dataset generated using iforest performs the best overall on the light gradient boosting machine model with an impactful accuracy of 0.985. Other anomaly detection technique-generated datasets also perform better than the initial dataset, with an accuracy of 0.859. The best model is validated using 10-fold validation.

4.1.1. Accuracy

The accuracy is calculated using Equation (5). Figure 4a shows the average accuracy of all the models. The light gradient boosting machine has the highest accuracy of 0.985, followed by xgboost and random forest. Extra trees have the lowest accuracy of 0.478.

4.1.2. F1 Score

The F1 is calculated using Equation (3). Figure 4b shows the average F1 score of all the models after 10-fold validation. The light gradient boosting machine has the highest F1 of 0.985, followed by xgboost and random forest. Extra trees have the lowest accuracy of 0.442.

4.1.3. MCC

The MCC is calculated using Equation (2). Figure 4c shows the average MCC of all the models after 10-fold validation. The light gradient boosting machine has the highest accuracy of 0.985, followed by xgboost and random forest. Extra trees have the lowest accuracy of 0.367.

4.2. K-Fold Validation

Cross-validation is a statistical method for estimating capacity in machine learning models. It is commonly used in applied machine learning to compare and select a model for a specific predictive modeling problem. It is easy to understand, implement, and produce skill estimates with a more negligible bias than other approaches. Cross-validation is a resampling technique used to assess ML models on a small sample of data. The process contains only one parameter called k, which specifies the number of groups in which a given data sample should be divided.
Consequently, this method is often referred to as k-fold cross-validation. When a specific number for k is specified, it may be used instead of k in the model reference, such as k = 10, resulting in a 10-cross-validation. Cross-validation is primarily used in applied machine learning to assess the competency of a machine learning model on unseen data. It uses a small sample to determine how the model will work when used to produce projections of data not used in model formation. The charts show that the built model is robust after 10 plies and is reliable for testing our sEMG data.

4.3. Results and Discussion

The results indicate that datasets created following the anomaly detection techniques increase the performance of classifiers for movement prediction.
The performance of each model is shown in Table 6. All the prediction models have been trained on default parameters and evaluated on MCC, accuracy, F1, precision, and AUC. The dataset has a small number of features but a large number of observation values. The K-fold cross-validation technique is used to assess the robustness of the best prediction method.
The dataset created with iforest achieves an outstanding overall score on the light gradient boosting machine model with an accuracy of 98.5% compared to the baseline performance without any adjustments on our dataset, which achieves an accuracy of 85.9%. The other models significantly improve their accuracy and the metrics after anomaly removal from the data. There is a performance increase of 13% accuracy from the models which do not use anomaly detection techniques. Compared to previous works [34,35] which were able to achieve an accuracy of 92.5% and 91%, this has also been improved by 6–7%, respectively, which proves that this technique is effective in boosting the accuracy for classification of sEMG signals. Thus, results demonstrate that eliminating anomalies substantially improves performance, producing some robust results, and should be used for developing models using sEMG data.

4.4. Limitation of the Study

Data imbalance is an essential issue in the classification of medical data. This can occur because of significant differences in the number of healthy and unhealthy individuals. This may happen because the data collection length varies depending on normal and abnormal subjects. An individual with an anomaly of the knee takes longer to accomplish the task of moving, resulting in a more extended signal. Therefore, the abnormal subject’s sEMG signal length results in a class imbalance. This affects the precision of the classification of knee motion and the prediction of anomalies.

5. Conclusions

This work demonstrated machine learning for movement classification using sEMG data. The suggested models are based on a large database with various parameters. This study aimed to analyze differences in classification performance following anomalous detection methods. It was used to identify events in EMG signals that differed significantly from most data and to create new data sets.
Real-time sEMG data analysis and movement classification can be beneficial in the medical field. Doctors could identify aberrant motions and provide preventative steps to minimize risks and injuries connected to joints or muscles. This work hopes to contribute towards a more accurate diagnosis system to predict abnormalities and suggest precautionary measures to prevent any unfortunate events. In future research, we plan to gather DGG data from other topics to reduce data imbalances, to test oversample methods, and study the most advanced machine learning graders. Convolution neural networks (CNN) have shown promising results in various classification questions. The iforest anomaly detection techniques presented here can be up to 98.5% accurate with light gradient boosting. It enhances the ability to detect defects within EMG signals. The proposed methodology can be used in muscular deficiency or any other physiological deficiency.

Author Contributions

Conceptualization, N.K.; Data curation, H.B. and B.C.; Formal analysis, B.C. and N.K.; Methodology, H.B., B.C., P.S.R. and N.K.; Validation, H.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Foundation, A. Arthritis by the Numbers: Book of Trusted Facts and Figures; Arthritis Foundation: Atlanta, GA, USA, 2017. [Google Scholar]
  2. Bedson, J.; Jordan, K.; Croft, P. How do GPs use x rays to manage chronic knee pain in the elderly? A case study. Ann. Rheum. Dis. 2003, 62, 450–454. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Kubakaddi, S.; Ravikumar, K.; Harini, D.G. Measurement of cartilage thickness for early detection of knee osteoarthritis (KOA). In Proceedings of the 2013 IEEE Point-of-Care Healthcare Technologies (PHT), Bangalore, India, 16–18 January 2013; pp. 208–211. [Google Scholar]
  4. Kim, Y.S.; Lee, J.; Lee, S.; Kim, M. A force reflected exoskeleton-type masterarm for human-robot interaction. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2005, 35, 198–212. [Google Scholar] [CrossRef]
  5. Islam, M.R.; Assad-Uz-Zaman, M.; Brahmi, B.; Bouteraa, Y.; Wang, I.; Rahman, M.H. Design and development of an upper limb rehabilitative robot with dual functionality. Micromachines 2021, 12, 870. [Google Scholar] [CrossRef]
  6. Zhang, L.; Liu, G.; Han, B.; Wang, Z.; Li, H.; Jiao, Y. Assistive devices of human knee joint: A review. Robot. Auton. Syst. 2020, 125, 103394. [Google Scholar] [CrossRef]
  7. Liu, D.X.; Xu, J.; Chen, C.; Long, X.; Tao, D.; Wu, X. Vision-assisted autonomous lower-limb exoskeleton robot. IEEE Trans. Syst. Man Cybern. Syst. 2019, 51, 3759–3770. [Google Scholar] [CrossRef]
  8. Sun, Y.; Tang, Y.; Zheng, J.; Dong, D.; Chen, X.; Bai, L. From sensing to control of lower limb exoskeleton: A systematic review. Annu. Rev. Control. 2022, 53, 83–96. [Google Scholar] [CrossRef]
  9. Calle-Siguencia, J.; Callejas-Cuervo, M.; García-Reino, S. Integration of Inertial Sensors in a Lower Limb Robotic Exoskeleton. Sensors 2022, 22, 4559. [Google Scholar] [CrossRef]
  10. Tang, X.; Wang, X.; Ji, X.; Zhou, Y.; Yang, J.; Wei, Y.; Zhang, W. A Wearable Lower Limb Exoskeleton: Reducing the Energy Cost of Human Movement. Micromachines 2022, 13, 900. [Google Scholar] [CrossRef]
  11. Taylor, C.R.; Srinivasan, S.S.; Yeon, S.H.; O’Donnell, M.; Roberts, T.; Herr, H.M. Magnetomi-crometry. Sci. Robot. 2021, 6, eabg0656. [Google Scholar] [CrossRef]
  12. Botter, A.; Vieira, T.; Carbonaro, M.; Cerone, G.L.; Hodson-Tole, E.F. Electrodes’ configuration influences the agreement between surface EMG and B-mode ultrasound detection of motor unit fasciculation. IEEE Access 2021, 9, 98110–98120. [Google Scholar] [CrossRef]
  13. Merletti, R.; De Luca, C.J. New techniques in surface electromyography. Comput. Aided Electromyogr. Expert Syst. 1989, 9, 115–124. [Google Scholar]
  14. Gui, K.; Liu, H.; Zhang, D. A practical and adaptive method to achieve EMG-based torque estimation for a robotic exoskeleton. IEEE/ASME Trans. Mechatron. 2019, 24, 483–494. [Google Scholar] [CrossRef]
  15. Kang, I.; Kunapuli, P.; Hsu, H.; Young, A.J. Electromyography (EMG) signal contributions in speed and slope estimation using robotic exoskeletons. In Proceedings of the 2019 IEEE 16th International Conference on Rehabilitation Robotics (ICORR), Toronto, ON, Canada, 24–28 June 2019; pp. 548–553. [Google Scholar]
  16. Kanoga, S.; Kanemura, A.; Asoh, H. Are armband sEMG devices dense enough for long-term 465 use?—Sensor placement shifts cause significant reduction in recognition accuracy. Biomed. Signal Process. Control. 2020, 60, 101981. [Google Scholar] [CrossRef]
  17. Rainoldi, A.; Melchiorri, G.; Caruso, I. A method for positioning electrodes during surface EMG recordings in lower limb muscles. J. Neurosci. Methods 2004, 134, 37–43. [Google Scholar] [CrossRef] [PubMed]
  18. Murley, G.S.; Menz, H.B.; Landorf, K.B.; Bird, A.R. Reliability of lower limb electromyography during overground walking: A comparison of maximal- and sub-maximal normalisation techniques. J. Biomech. 2010, 43, 749–756. [Google Scholar] [CrossRef]
  19. Mulroy, S.; Gronley, J.; Weiss, W.; Newsam, C.; Perry, J. Use of cluster analysis for gait pattern classification of patients in the early and late recovery phases following stroke. Gait Posture 2003, 18, 114–125. [Google Scholar] [CrossRef]
  20. Huang, H.; Zhang, F.; Hargrove, L.J.; Dou, Z.; Rogers, D.R.; Englehart, K.B. Continuous 474 locomotion-mode identification for prosthetic legs based on neuromuscular–mechanical fusion. IEEE Trans. Biomed. Eng. 2011, 58, 2867–2875. [Google Scholar] [CrossRef] [Green Version]
  21. Chen, J.; Zhang, X.; Cheng, Y.; Xi, N. Surface EMG based continuous estimation of human lower limb joint angles by using deep belief networks. Biomed. Signal Process. Control. 2018, 40, 335–342. [Google Scholar] [CrossRef]
  22. Morbidoni, C.; Cucchiarelli, A.; Fioretti, S.; Di Nardo, F. A Deep Learning Approach to EMG-Based Classification of Gait Phases during Level Ground Walking. Electronics 2019, 8, 894. [Google Scholar] [CrossRef] [Green Version]
  23. Bonato, P.; Heng, M.; Gonzalez-Cueto, J.; Leardini, A.; O’Connor, J.; Roy, S.H. EMG-based 481 measures of fatigue during a repetitive squat exercise. IEEE Eng. Med. Biol. Mag. 2001, 20, 133–143. [Google Scholar] [CrossRef]
  24. Takhar, G.; Prakash, C.; Mittal, N.; Kumar, R. Vision-based gender recognition using hybrid background subtraction technique. In Proceedings of the International Conference on Next Generation Computing Technologies, Dehradun, India, 30–31 October 2017; pp. 651–662. [Google Scholar]
  25. EMG Dataset in Lower Limb Data Set. Available online: http://archive.ics.uci.edu/ml/datasets/emg+dataset+in+lower+limb (accessed on 28 July 2022).
  26. Khaledian, E.; Pandey, S.; Kundu, P.; Srivastava, A. Real-time Synchrophasor Data Anomaly Detection and Classification using Isolation Forest, KMeans and LoOP. IEEE Trans. Smart Grid 2020, 12, 2378–2388. [Google Scholar] [CrossRef]
  27. Aimal, S.; Javaid, N.; Islam, T.; Khan, W.Z.; Aalsalem, M.Y.; Sajjad, H. An efficient cnn 492 and knn data analytics for electricity load forecasting in the smart grid. In Proceedings of the Workshops of the International Conference on Advanced Information Networking and Applications, Matsue, Japan, 27–29 March 2019; pp. 592–603. [Google Scholar]
  28. Hurst, W.; Montanez, C.A.C.; Shone, N. Time-Pattern Profiling from Smart Meter Data to Detect Outliers in Energy Consumption. IoT 2020, 1, 92–108. [Google Scholar] [CrossRef]
  29. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. In Proceedings of the 31st Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  30. Chen, T.; He, T.; Benesty, M.; Khotilovich, V.; Tang, Y.; Cho, H.; Chen, K. Xgboost: Extreme gradient boosting. R Package Version 2015, 1, 1–4. [Google Scholar]
  31. Biau, G.; Scornet, E. A random forest guided tour. TEST 2016, 25, 197–227. [Google Scholar] [CrossRef] [Green Version]
  32. Ahmad, M.W.; Reynolds, J.; Rezgui, Y. Predictive modelling for solar thermal energy systems: A comparison of support vector regression, random forest, extra trees and regression trees. J. Clean. Prod. 2018, 203, 810–821. [Google Scholar] [CrossRef]
  33. Myles, A.J.; Feudale, R.N.; Liu, Y.; Woody, N.A.; Brown, S.D. An introduction to decision tree modeling. J. Chemom. A J. Chemom. Soc. 2004, 18, 275–285. [Google Scholar] [CrossRef]
  34. Vijayvargiya, A.; Prakash, C.; Kumar, R.; Bansal, S.; Tavares, J.M.R.S. Human knee abnormality detection from imbalanced sEMG data. Biomed. Signal Process. Control. 2021, 66, 102406. [Google Scholar] [CrossRef]
  35. Vijayvargiya, A.; Kumar, R.; Dey, N.; Tavares, J.M.R. Comparative analysis of machine learning techniques for the classification of knee abnormality. In Proceedings of the 2020 IEEE 5th International Conference on Computing Communication and Automation (ICCCA), Greater Noida, India, 30–31 October 2020; pp. 1–6. [Google Scholar]
Figure 1. Muscle diagram for rectus femoris (RF), vastus lateralis (VL), vastus medialis (VM), semitendinosus (ST), biceps femoris (BF), and semitmembranosus (SM).
Figure 1. Muscle diagram for rectus femoris (RF), vastus lateralis (VL), vastus medialis (VM), semitendinosus (ST), biceps femoris (BF), and semitmembranosus (SM).
Sustainability 14 13464 g001
Figure 2. The sEMG signals at various stages during data preprocessing.
Figure 2. The sEMG signals at various stages during data preprocessing.
Sustainability 14 13464 g002
Figure 3. Methodology used.
Figure 3. Methodology used.
Sustainability 14 13464 g003
Figure 4. Cross-validation of best model for accuracy, F1-score, and MCC.
Figure 4. Cross-validation of best model for accuracy, F1-score, and MCC.
Sustainability 14 13464 g004
Table 1. Description of the dataset.
Table 1. Description of the dataset.
DescriptionType/Count
Number of features5
Feature characteristicsReal
No. of classes (Movements)6
Dataset characteristicsMultivariate
Associated tasksClassification
Missing valuesNo
Table 2. Description of the features.
Table 2. Description of the features.
FeatureInformation
RFRectus Femoris
BFBiceps Femoris
VMVastus Medialis
STSemitendinosus
FXFlexion
Table 3. Sample dataset.
Table 3. Sample dataset.
RFBFVMSTFXMovement
0.0011−0.00110.00210.005057.60
0.0011−0.00100.00220.005257.50
0.0011−0.00090.00240.005457.30
0.00510.0002−0.00100.000731
0.00520.0002−0.00070.000731
0.00560.0003−0.00010.000731
0.00110.00020.00020.000070.12
0.00110.00020.00030.000070.12
0.00110.00020.00040.000070.12
−0.00040.00020.01840.0017−93
0.00040.00110.01870.0022−93
0.00120.00210.01900.0028−93
0.00110.00180.00080.0012−4.64
0.00110.00190.00090.0013−4.54
0.00120.00200.00090.0013−4.64
0.00430.0013−0.00190.0005−305
0.00450.0013−0.00170.0005−29.75
0.00480.0014−0.00150.0005−29.45
Table 4. Statistical analysis of each feature in each dataset.
Table 4. Statistical analysis of each feature in each dataset.
DatasetFeaturesMean.Std.Min.Max.
RF0.040.016−0.110.10
BF−0.020.008−0.060.11
Nor.WalkVM−0.010.026−0.180.17
ST−0.030.011−0.080.11
FX19.3617.8361.3062.60
RF−0.010.083−1.271.36
BF−0.060.023−0.320.16
Abn.WalkVM−0.060.068−0.610.59
ST0.060.016−0.150.10
FX−12.716.37−57.23.10
RF0.030.016−0.110.10
BF−0.020.008−0.060.11
Nor.SitVM−0.010.026−0.180.17
ST−0.030.011−0.080.11
FX53.3636.8363.30109.6
RF−0.030.006−0.080.05
BF−0.060.034−0.330.20
Abn.SitVM−0.040.012−0.110.20
ST0.050.06−0.560.40
FX−58.744.37−113−2.90
RF0.030.018−0.110.10
BF−0.030.003−0.060.11
Nor.StandVM−0.010.026−0.180.17
ST−0.040.011−0.080.11
FX51.3620.83620.3075.6
RF−0.010.032−0.200.26
BF−0.060.006−0.050.06
Abn.StandVM−0.060.019−0.160.14
ST0.070.02−0.490.35
FX−29.430.37−83−4.70
Table 5. The importance of each feature.
Table 5. The importance of each feature.
RunsFXBFSTVMRF
10.250.220.20.190.14
20.230.20.240.170.16
30.270.230.190.160.15
40.250.210.210.170.16
50.240.20.20.170.17
Avg.0.2480.2120.2080.1720.156
Table 6. Performance comparison of different anomaly detection techniques with different prediction models on different metrices on all the datasets.
Table 6. Performance comparison of different anomaly detection techniques with different prediction models on different metrices on all the datasets.
Anomaly Detection
Technique
Models lgbmxgboostrfdtet
Metrices ↓
Without Anomaly DetectionMCC0.830.7530.5590.5370.367
AUC0.9720.9580.8970.8310.839
Precision0.8590.7980.6130.5140.59
F10.8570.790.6170.5220.442
Accuracy0.8590.7970.640.5370.478
iforestMCC0.9850.9820.9220.8220.816
AUC0.9860.9860.9830.9670.974
Precision0.9850.9840.9370.8590.865
F10.9850.9830.9330.8550.843
Accuracy0.9850.9830.9350.8560.852
KNNMCC0.9220.90.7850.5690.72
AUC0.9830.980.9480.8620.918
Precision0.9560.9460.8930.80.861
F10.9540.9420.8840.7650.854
Accuracy0.9530.9410.880.7480.853
lofMCC0.9080.8050.5580.4540.43
AUC0.9790.9620.8940.8370.834
Precision0.9260.8470.6920.60.631
F10.9260.8470.6530.5710.506
Accuracy0.9260.8470.6660.5830.569
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bansal, H.; Chinagundi, B.; Rana, P.S.; Kumar, N. An Ensemble Machine Learning Technique for Detection of Abnormalities in Knee Movement Sustainability. Sustainability 2022, 14, 13464. https://doi.org/10.3390/su142013464

AMA Style

Bansal H, Chinagundi B, Rana PS, Kumar N. An Ensemble Machine Learning Technique for Detection of Abnormalities in Knee Movement Sustainability. Sustainability. 2022; 14(20):13464. https://doi.org/10.3390/su142013464

Chicago/Turabian Style

Bansal, Hunish, Basavraj Chinagundi, Prashant Singh Rana, and Neeraj Kumar. 2022. "An Ensemble Machine Learning Technique for Detection of Abnormalities in Knee Movement Sustainability" Sustainability 14, no. 20: 13464. https://doi.org/10.3390/su142013464

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop